Skip to content

Conversation

ikawrakow
Copy link
Contributor

I have missed this tweak when adding Q2_K_S.

With this change, model size for Mistral-7B increases by only ~30 MB (0.03 bpw) while

  • Perplexity for a context of 512 on wiki.test.raw goes down from 6.9259 to 6.7116
  • 10-shot HellaSwag score after 2000 tasks increases by 0.95 +/- 0.42.

@ggerganov ggerganov merged commit 2b3a665 into master Jan 17, 2024
jordankanter pushed a commit to jordankanter/llama.cpp that referenced this pull request Feb 3, 2024
hodlen pushed a commit to hodlen/llama.cpp that referenced this pull request Apr 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants